Joint Speed Control and Energy Replenishment Optimization for UAV-Assisted IoT Data Collection With Deep Reinforcement Transfer Learning

نویسندگان

چکیده

Unmanned-aerial-vehicle (UAV)-assisted data collection has been emerging as a prominent application due to its flexibility, mobility, and low operational cost. However, under the dynamic uncertainty of Internet Things energy replenishment processes, optimizing performance for UAV collectors is very challenging task. Thus, this article introduces novel framework that jointly optimizes flying speed each significantly improve overall system (e.g., usage efficiency). Specifically, we first develop Markov decision process help automatically dynamically make optimal decisions dynamics uncertainties environment. Although traditional reinforcement learning algorithms, such $Q$ -learning deep -learning, can obtain policy, they often take long time converge require high computational complexity. Therefore, it impractical deploy these conventional methods on UAVs with limited computing capacity resource. To end, advanced transfer techniques allow “share” “transfer” knowledge, thereby reducing well improving quality. Extensive simulations demonstrate our proposed solution average up 200% reduce convergence 50% compared those methods.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning State Representations for Query Optimization with Deep Reinforcement Learning

Deep reinforcement learning is quickly changing the field of artificial intelligence. These models are able to capture a high level understanding of their environment, enabling them to learn difficult dynamic tasks in a variety of domains. In the database field, query optimization remains a difficult problem. Our goal in this work is to explore the capabilities of deep reinforcement learning in...

متن کامل

Reinforcement Learning with Deep Energy-Based Policies

We propose a method for learning expressive energy-based policies for continuous states and actions, which has been feasible only in tabular domains before. We apply our method to learning maximum entropy policies, resulting into a new algorithm, called soft Q-learning, that expresses the optimal policy via a Boltzmann distribution. We use the recently proposed amortized Stein variational gradi...

متن کامل

On-line Building Energy Optimization using Deep Reinforcement Learning

Unprecedented high volumes of data are becoming available with the growth of the advanced metering infrastructure. These are expected to benefit planning and operation of the future power system, and to help the customers transition from a passive to an active role. In this paper, we explore for the first time in the smart grid context the benefits of using Deep Reinforcement Learning, a hybrid...

متن کامل

Simulated Transfer Learning Through Deep Reinforcement Learning

This paper encapsulates the use reinforcement learning on raw images provided by a simulation to produce a partially trained network. Before training is continued, this partially trained network is fed different raw images that are more tightly coupled with a richer representation of the non-simulated environment. The use of transfer learning allows for the model to adjust to this richer repres...

متن کامل

Reinforcement Learning Based PID Control of Wind Energy Conversion Systems

In this paper an adaptive PID controller for Wind Energy Conversion Systems (WECS) has been developed. Theadaptation technique applied to this controller is based on Reinforcement Learning (RL) theory. Nonlinearcharacteristics of wind variations as plant input, wind turbine structure and generator operational behaviordemand for high quality adaptive controller to ensure both robust stability an...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Internet of Things Journal

سال: 2023

ISSN: ['2372-2541', '2327-4662']

DOI: https://doi.org/10.1109/jiot.2022.3151201